RFpredInterval: An R Package for Prediction Intervals with Random Forests and Boosted Forests
نویسندگان
چکیده
Like many predictive models, random forests provide point predictions for new observations. Besides the prediction, it is important to quantify uncertainty in prediction. Prediction intervals information about reliability of predictions. We have developed a comprehensive R package, [RFpredInterval](https://CRAN.R-project.org/package=RFpredInterval), that integrates 16 methods build prediction with and boosted forests. The set implemented package includes method (PIBF) 15 variations produce forests, as proposed by [@roy_prediction_2020]. perform an extensive simulation study apply real data analyses compare performance ten existing building results show very competitive and, globally, outperforms competing methods.
منابع مشابه
VSURF: An R Package for Variable Selection Using Random Forests
This paper describes the R package VSURF. Based on random forests, and for both regression and classification problems, it returns two subsets of variables. The first is a subset of important variables including some redundancy which can be relevant for interpretation, and the second one is a smaller subset corresponding to a model trying to avoid redundancy focusing more closely on the predict...
متن کاملConfidence Intervals for Random Forests Confidence Intervals for Random Forests: The Jackknife and the Infinitesimal Jackknife
We study the variability of predictions made by bagged learners and random forests, and show how to estimate standard errors for these methods. Our work builds on variance estimates for bagging proposed by Efron (1992, 2012) that are based on the jackknife and the infinitesimal jackknife (IJ). In practice, bagged predictors are computed using a finite number B of bootstrap replicates, and worki...
متن کاملRandom Forests for Ordinal Response Data: Prediction and Variable Selection
The random forest method is a commonly used tool for classification with high-dimensional data that is able to rank candidate predictors through its inbuilt variable importance measures (VIMs). It can be applied to various kinds of regression problems including nominal, metric and survival response variables. While classification and regression problems using random forest methodology have been...
متن کاملRandom Prism: An Alternative to Random Forests
Ensemble learning techniques generate multiple classifiers, so called base classifiers, whose combined classification results are used in order to increase the overall classification accuracy. In most ensemble classifiers the base classifiers are based on the Top Down Induction of Decision Trees (TDIDT) approach. However, an alternative approach for the induction of rule based classifiers is th...
متن کاملMondrian Forests: Efficient Online Random Forests
Ensembles of randomized decision trees, usually referred to as random forests, are widely used for classification and regression tasks in machine learning and statistics. Random forests achieve competitive predictive performance and are computationally efficient to train and test, making them excellent candidates for real-world prediction tasks. The most popular random forest variants (such as ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: R Journal
سال: 2022
ISSN: ['2073-4859']
DOI: https://doi.org/10.32614/rj-2022-012